417 research outputs found

    Invariance Matters: Exemplar Memory for Domain Adaptive Person Re-identification

    Full text link
    This paper considers the domain adaptive person re-identification (re-ID) problem: learning a re-ID model from a labeled source domain and an unlabeled target domain. Conventional methods are mainly to reduce feature distribution gap between the source and target domains. However, these studies largely neglect the intra-domain variations in the target domain, which contain critical factors influencing the testing performance on the target domain. In this work, we comprehensively investigate into the intra-domain variations of the target domain and propose to generalize the re-ID model w.r.t three types of the underlying invariance, i.e., exemplar-invariance, camera-invariance and neighborhood-invariance. To achieve this goal, an exemplar memory is introduced to store features of the target domain and accommodate the three invariance properties. The memory allows us to enforce the invariance constraints over global training batch without significantly increasing computation cost. Experiment demonstrates that the three invariance properties and the proposed memory are indispensable towards an effective domain adaptation system. Results on three re-ID domains show that our domain adaptation accuracy outperforms the state of the art by a large margin. Code is available at: https://github.com/zhunzhong07/ECNComment: To appear in CVPR 201

    GridNet with automatic shape prior registration for automatic MRI cardiac segmentation

    Full text link
    In this paper, we propose a fully automatic MRI cardiac segmentation method based on a novel deep convolutional neural network (CNN) designed for the 2017 ACDC MICCAI challenge. The novelty of our network comes with its embedded shape prior and its loss function tailored to the cardiac anatomy. Our model includes a cardiac centerof-mass regression module which allows for an automatic shape prior registration. Also, since our method processes raw MR images without any manual preprocessing and/or image cropping, our CNN learns both high-level features (useful to distinguish the heart from other organs with a similar shape) and low-level features (useful to get accurate segmentation results). Those features are learned with a multi-resolution conv-deconv "grid" architecture which can be seen as an extension of the U-Net. Experimental results reveal that our method can segment the left and right ventricles as well as the myocardium from a 3D MRI cardiac volume in 0.4 second with an average Dice coefficient of 0.90 and an average Hausdorff distance of 10.4 mm.Comment: 8 pages, 1 tables, 2 figure

    Ground state solutions for diffusion system with superlinear nonlinearity

    Get PDF
    In this paper, we study the following diffusion system \begin{equation*} \begin{cases} \partial_{t}u-\Delta_{x} u +b(t,x)\cdot \nabla_{x} u +V(x)u=g(t,x,v),\\ -\partial_{t}v-\Delta_{x} v -b(t,x)\cdot \nabla_{x} v +V(x)v=f(t,x,u) \end{cases} \end{equation*} where z=(u,v) ⁣:R×RNR2z=(u,v)\colon\mathbb{R}\times\mathbb{R}^{N}\rightarrow\mathbb{R}^{2}, bC1(R×RN,RN)b\in C^{1}(\mathbb{R}\times\mathbb{R}^{N}, \mathbb{R}^{N}) and V(x)C(RN,R)V(x)\in C(\mathbb{R}^{N},\mathbb{R}). Under suitable assumptions on the nonlinearity, we establish the existence of ground state solutions by the generalized Nehari manifold method developed recently by Szulkin and Weth

    Detect Any Deepfakes: Segment Anything Meets Face Forgery Detection and Localization

    Full text link
    The rapid advancements in computer vision have stimulated remarkable progress in face forgery techniques, capturing the dedicated attention of researchers committed to detecting forgeries and precisely localizing manipulated areas. Nonetheless, with limited fine-grained pixel-wise supervision labels, deepfake detection models perform unsatisfactorily on precise forgery detection and localization. To address this challenge, we introduce the well-trained vision segmentation foundation model, i.e., Segment Anything Model (SAM) in face forgery detection and localization. Based on SAM, we propose the Detect Any Deepfakes (DADF) framework with the Multiscale Adapter, which can capture short- and long-range forgery contexts for efficient fine-tuning. Moreover, to better identify forged traces and augment the model's sensitivity towards forgery regions, Reconstruction Guided Attention (RGA) module is proposed. The proposed framework seamlessly integrates end-to-end forgery localization and detection optimization. Extensive experiments on three benchmark datasets demonstrate the superiority of our approach for both forgery detection and localization. The codes will be released soon at https://github.com/laiyingxin2/DADF

    Analyse de trafic routier à partir de vidéos à faible débit

    Get PDF
    Abstract: Nowadays, traffic analysis are relying on data collected from various traffic sensors. Among the various traffic surveillance techniques, video surveillance systems are often used for monitoring and characterizing traffic load. In this thesis, we focused on two aspects of traffic analysis without using motion features in low frame-rate videos: Traffic density flow analysis and Vehicle detection and classification. Traffic density flow analysis}: Knowing in real time when the traffic is fluid or when it jams is a key information to help authorities re-route vehicles and reduce congestion. Accurate and timely traffic flow information is strongly needed by individual travelers, the business sectors and government agencies. In this part, we investigated the possibility of monitoring highway traffic based on videos whose frame rate is too low to accurately estimate motion features. As we are focusing on analyzing traffic images and low frame-rate videos, traffic density is defined as the percentage of road being occupied by vehicles. In our previous work, we validated that traffic status is highly correlated to its texture features and that Convolutional Neural Networks (CNN) has the superiority of extracting discriminative texture features. We proposed several CNN models to segment traffic images into three different classes (road, car and background), classify traffic images into different categories (empty, fluid, heavy, jam) and predict traffic density without using any motion features. In order to generalize the model trained on a specific dataset to analyze new traffic scenes, we also proposed a novel transfer learning framework to do model adaptation. Vehicle detection and classification: The detection of vehicles pictured by traffic cameras is often the very first step of video surveillance systems, such as vehicle counting, tracking and retrieval. In this part, we explore different deep learning methods applied to vehicle detection and classification. Firstly, realizing the importance of large dataset for traffic analysis, we built and released the largest traffic dataset (MIO-TCD) in the world for vehicle localization and classification in collaboration with colleagues from Miovision inc. (Waterloo, On). With this dataset, we organized the Traffic Surveillance Workshop and Challenge in conjunction with CVPR 2017. Secondly, we evaluated several state-of-the-art deep learning methods for the classification and localization task on the MIO-TCD dataset. In light of the results, we may conclude that state-of-the-art deep learning methods exhibit a capacity to localize and recognize vehicle from a single video frame. While with a deep analysis of the results, we also identify scenarios for which state-of-the-art methods are still failing and propose concrete ideas for future work. Lastly, as saliency detection aims to highlight the most relevant objects in an image (e.g. vehicles in traffic scenes), we proposed a multi-resolution 4*5 grid CNN model for the salient object detection. The model enables near real-time high performance saliency detection. We also extend this model to do traffic analysis, experiment results show that our model can precisely segment foreground vehicles in traffic scenes.De nos jours, l’analyse de trafic routier est de plus en plus automatisée et s’appuie sur des données issues de senseurs en tout genre. Parmi les approches d’analyse de trafic routier figurent les méthodes à base de vidéo. Les méthodes à base de vidéo ont pour but d’identifier et de reconnaître les objets en mouvement (généralement des voitures et des piétons) et de comprendre leur dynamique. Un des défis parmi les plus difficile à résoudre est d’analyser des séquences vidéo dont le nombre d’images par seconde est très faible. Ce type de situation est pourtant fréquent considérant qu’il est très difficile (voir impossible) de transmettre et de stocker sur un serveur un très grand nombre d’images issues de plusieurs caméras. Dans ce cas, les méthodes issues de l’état de l’art échouent car un faible nombre d’images par seconde ne permet pas d’extraire les caractéristiques vidéos utilisées par ces méthodes tels le flux optique, la détection de mouvement et le suivi de véhicules. Au cours de cette thèse, nous nous sommes concentré sur l’analyse de trafic routier à partir de séquences vidéo contenant un très faible nombre d’images par seconde. Plus particulièrement, nous nous sommes concentrés sur les problème d’estimation de la densité du trafic routier et de la classification de véhicules. Pour ce faire, nous avons proposé différents modèles à base de réseaux de neurones profonds (plus particulièrement des réseaux à convolution) ainsi que de nouvelles bases de données permettant d’entraîner les dits modèles. Parmi ces bases de données figure « MIO-TCD », la plus grosse base de données annotées au monde faite pour l’analyse de trafic routier

    Assessment of Environmental Performance of Rapid Prototyping and Rapid Tooling Processes

    Get PDF
    A method for assessing the environmental performance of solid freeform fabrication (SFF) based rapid prototyping and rapid tooling processes is presented in this paper. In this method of assessment, each process is divided into a number of life stages. The environmental effect of each process stage is analyzed and evaluated based on an environmental index utilizing the Eco-indicators that were compiled by PreConsultants of the Netherlands. The effects of various life stages are then combined to obtain the environmental performance of a process. In the assessment of SFF processes, we consider the material use in the fabrication of a part, energy consumption, process wastes, and disposal of a part after its normal life. An example is given to illustrate this assessment method applied to the stereolithography (SLA) process and two SLA based rapid tooling processes

    Environmental Performance Analysis of Solid Freedom Fabrication Processes

    Get PDF
    This paper presents a method for analyzing the environmental performance of solid freeform fabrication (SFF) processes. In this method, each process is divided into life phases. Environmental effects of every process phase are then analyzed and evaluated based on the environmental and resource management data. These effects are combined to obtain the environmental performance of the process. The analysis of the environmental performance of SFF processes considers the characteristics of SFF technology, includes material, energy consumption, processes wastes, and disposal. Case studies for three typical SFF processes: stereolithography (SL); selective laser sintering (SLS); and fused deposition modeling (FDM) are presented to illustrate this method

    Lifecycle Analysis for Environmentally Conscious Solid Freeform Manufacturing

    Get PDF
    A lifecycle based process model for analyzing the environmental performance of SFM processes and SFM based rapid tooling processes is presented in this paper. The process environmental performance assessment model considers material, energy and disposal scenarios. The material use, process parameters (e.g. scanning speed) and power use can affect the environmental consequence of a process when material resource, energy, human health and environmental damage are taken into account. The presented method is applied to the SLA process and two SLA based rapid tooling processes. The method can be used to compare different rapid prototyping (RP) and RT processes in terms of their environmental friendliness and for further multi-objective decision makin
    corecore